Regularization, Boosting and Mirror Averaging

نویسندگان

  • ALEXANDRE TSYBAKOV
  • Peter Bickel
  • Bo Li
چکیده

In their paper, Peter Bickel and Bo Li give an interesting unified view of regularization methods in statistics. The literature on this subject is immense, so they outline a general conceptual approach, and then focus on some selected problems where regularization is used, such as regression and classification, or more generally, prediction. In this context, they discuss in detail a number of recently emerging techniques, in particular, boosting, estimation of large covariance matrices, estimation in the models where the dimension is larger than the sample size. It is difficult to overestimate the importance of regularization in statistics, especially in nonparametrics. Most of nonparametric estimation problems are ill-posed, and common estimators (kernel, histogram, spline, orthogonal series etc.) are nothing but regularized methods of solving them. The corresponding regularization parameters are just smoothing parameters of the estimators. The main ideas of statistical regularization can be very transparently explained for prediction problems. Assume that X1, . . . , Xn are i.i.d. observations taking values in a space X , and assume that the unknown underlying function f ? that we want to estimate belongs to a space F . Consider a loss function Q : X × F → IR and the associated prediction risk R(f) = EQ(X, f) where X has the same distribution as Xi. Assume that f ? is a minimizer of the risk R(f) over F . Then a classical, but not always reasonable, estimator of f ? is a minimizer over f ∈ F of the corresponding empirical risk

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Unified View of Regularized Dual Averaging and Mirror Descent with Implicit Updates

We study three families of online convex optimization algorithms: follow-the-proximallyregularized-leader (FTRL-Proximal), regularized dual averaging (RDA), and compositeobjective mirror descent. We first prove equivalence theorems that show all of these algorithms are instantiations of a general FTRL update. This provides theoretical insight on previous experimental observations. In particular...

متن کامل

Follow-the-Regularized-Leader and Mirror Descent: Equivalence Theorems and L1 Regularization

We prove that many mirror descent algorithms for online convex optimization (such as online gradient descent) have an equivalent interpretation as follow-the-regularizedleader (FTRL) algorithms. This observation makes the relationships between many commonly used algorithms explicit, and provides theoretical insight on previous experimental observations. In particular, even though the FOBOS comp...

متن کامل

EM and component-wise boosting for Hidden Markov Models: a machine-learning approach to capture-recapture

1 This study presents a new boosting method for capture-recapture models, rooted in predictive2 performance and machine-learning. The regularization algorithm combines Expectation-Maximization and 3 boosting to yield a type of multimodel inference, including automatic variable selection and control of model 4 complexity. By analyzing simulations and a real dataset, this study shows the qualitat...

متن کامل

Topics in Regularization and Boosting

Regularization is critical for successful statistical modeling of “modern” data, which is high-dimensional, sometimes noisy and often contains a lot of irrelevant predictors. It exists — implicitly or explicitly — at the heart of all successful methods. The two main challenges which we take on in this thesis are understanding its various aspects better and suggesting new regularization approach...

متن کامل

Boosting Algorithms: Regularization, Prediction and Model Fitting

We present a statistical perspective on boosting. Special emphasis is given to estimating potentially complex parametric or nonparametric models, including generalized linear and additive models as well as regression models for survival analysis. Concepts of degrees of freedom and corresponding Akaike or Bayesian information criteria, particularly useful for regularization and variable selectio...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2007